Mine rescue robots operate in extremely adverse subterranean environments, where the acquired video data are frequently affected by severe jitter and motion distortion. Such instability leads to the loss of critical visual information, thereby reducing the reliability of rescue decision-making. To address this issue, a dual-channel visual stabilization framework based on the SuperPoint network is proposed, extending the traditional ORB descriptor framework. Here, dual-channel refers to two configurable and mutually exclusive feature extraction paths—an ORB-based path and a SuperPoint-based path—that can be flexibly switched according to scene conditions and computational requirements, rather than operating simultaneously on the same frame. The subsequent stabilization pipeline remains unified and consistent across both modes. The method employs an optimized detector head that integrates deep feature extraction, non-maximum suppression, and boundary filtering to enable precise estimation of inter-frame motion. When combined with smoothing filters, the approach effectively attenuates vibrations induced by irregular terrain and dynamic operational conditions. Experimental evaluations conducted across diverse scenarios demonstrate that the proposed algorithm achieves an average improvement of 27.91% in Peak Signal-to-Noise Ratio (PSNR), a 55.04% reduction in Mean Squared Error (MSE), and more than a twofold increase in the Structural Similarity Index (SSIM) relative to pre-stabilized sequences. Moreover, runtime analysis indicates that the algorithm can operate in near-real-time, supporting its practical deployment on embedded mine rescue robot platforms.These results verify the algorithm’s robustness and applicability in environments requiring high visual stability and image fidelity, providing a reliable foundation for enhanced visual perception and autonomous decision-making in complex disaster scenarios.
Loading....